DLG Section: Experimentation design: Part 2
📄

DLG Section: Experimentation design: Part 2

Great! We now have a good enough understanding of where to experiment and the variations for experimentation. Now, let's understand who to experiment with and finally implement it.

1. Problem statement & overview
2. Hypothesis, goal setting & success metrics
3. Experimentation design: Part 1
4. Experimentation design: Part 2
5. Post experiment learnings & next steps
6. Stakeholder management


Audience & sample size

Define exactly which audience segment you will be testing with.

Key Considerations:

  • Who will participate in the experiment? (e.g., new users, returning customers, high-intent visitors)
  • How will the audience be selected? (e.g., organic traffic, paid campaigns, segmentation by geography or behaviour)
  • Are there any exclusion criteria? (e.g., users who have already interacted with a similar feature)

Determining Sample Size

The sample size should be large enough to ensure statistical significance. A small sample may lead to inconclusive results, while a large sample may require unnecessary resources.

Steps to Determine Sample Size:

  1. Identify baseline conversion rates.
  2. Use a sample size calculator (e.g., Optimizely, SurveyMonkey) to determine the required sample size.

For example;


Audience Segment:

Incoming organic traffic will be considered to remove any bias


Sample size:

Control: 500

Variation A: 300

Variation B: 300


Duration of the test:

March 30th to April 12th or till we’re statistically significant; whichever is earlier

Online sample size calculator


Optimizely

SurveyMonkey



Implementation and A/A Test

Implementation Plan

Successful experiment implementation requires a structured approach. Each experiment should have:

  • A clear experiment setup documented.
  • Necessary engineering and design changes planned.
  • Tracking and analytics correctly implemented.

Key Implementation Steps:

  1. Set up the test within an experimentation platform (e.g., Google Optimize, Optimizely, VWO).
  2. Ensure proper event tracking via analytics tools (e.g., Google Analytics, Mixpanel, Amplitude).
  3. Test the setup before launching to detect any inconsistencies.
  4. Deploy the experiment gradually if necessary (e.g., rolling out to 10% of traffic first).

A/A Test for Validation

An A/A test is conducted before launching an A/B test to validate the experiment setup. In an A/A test, run two identical versions of an experiment to baseline against each other. If the results are not the same, there is something wrong with the experiment setup. This could be right from the instrumentation to the way the users have been distributed.







Brand focused courses

Great brands aren't built on clicks. They're built on trust. Craft narratives that resonate, campaigns that stand out, and brands that last.

View all courses

All courses

Master every lever of growth — from acquisition to retention, data to events. Pick a course, go deep, and apply it to your business right away.

View all courses

Explore foundations by GrowthX

Built by Leaders From Amazon, CRED, Zepto, Hindustan Unilever, Flipkart, paytm & more

View All Foundations

Crack a new job or a promotion with the Career Centre

Designed for mid-senior & leadership roles across growth, product, marketing, strategy & business

View All Resources

Learning Resources

Browse 500+ case studies, articles & resources the learning resources that you won't find on the internet.

Patience—you’re about to be impressed.